A New Framework for Sign Language Recognition based on 3D Handshape Identification and Linguistic Modeling
نویسندگان
چکیده
Current approaches to sign recognition by computer generally have at least some of the following limitations: they rely on laboratory conditions for sign production, are limited to a small vocabulary, rely on 2D modeling (and therefore cannot deal with occlusions and off-plane rotations), and/or achieve limited success. Here we propose a new framework that (1) provides a new tracking method less dependent than others on laboratory conditions and able to deal with variations in background and skin regions (such as the face, forearms, or other hands); (2) allows for identification of 3D hand configurations that are linguistically important in American Sign Language (ASL); and (3) incorporates statistical information reflecting linguistic constraints in sign production. For purposes of large-scale computer-based sign language recognition from video, the ability to distinguish hand configurations accurately is critical. Our current method estimates the 3D hand configuration to distinguish among 77 hand configurations linguistically relevant for ASL. Constraining the problem in this way makes recognition of 3D hand configuration more tractable and provides the information specifically needed for sign recognition. Further improvements are obtained by incorporation of statistical information about linguistic dependencies among handshapes within a sign derived from an annotated corpus of almost 10,000 sign tokens.
منابع مشابه
The Importance of 3D Motion Trajectories for Computer-based Sign Recognition
Computer-based sign language recognition from video is a challenging problem because of the spatiotemporal complexities inherent in sign production and the variations within and across signers. However, linguistic information can help constrain sign recognition to make it a more feasible classification problem. We have previously explored recognition of linguistically significant 3D hand config...
متن کاملHand Tracking and Affine Shape-Appearance Handshape Sub-units in Continuous Sign Language Recognition
We propose and investigate a framework that utilizes novel aspects concerning probabilistic and morphological visual processing for the segmentation, tracking and handshape modeling of the hands, which is used as front-end for sign language video analysis. Our ultimate goal is to explore the automatic Handshape Sub-Unit (HSU) construction and moreover the exploitation of the overall system in a...
متن کاملAdvances in Dynamic-Static Integration of Movement and Handshape Cues for Sign Language Recognition
We explore the integration of movement-position (MP) and handshape (HS) cues for sign language recognition. The proposed method combines the data-driven subunit (SU) modeling exploiting the dynamicstatic notion for MP and the affine shape-appearance SUs for HS configurations. These aspects lead to the new dynamic-static integration of manual cues. This data-driven scheme takes advantage of the ...
متن کاملInfluence of Handshape Information on Automatic Sign Language Recognition
Research on automatic sign language recognition (ASLR) has mostly been conducted from a machine learning perspective. We propose to implement results from human sign recognition studies in ASLR. In a previous study it was found that handshape is important for human sign recognition. The current paper describes the implementation of this conclusion: using handshape in ASLR. Handshape information...
متن کاملDynamic Affine-Invariant Shape-Appearance Handshape Features and Classification in Sign Language Videos
We propose the novel approach of dynamic affine-invariant shape-appearance model (Aff-SAM) and employ it for handshape classification and sign recognition in sign language (SL) videos. AffSAM offers a compact and descriptive representation of hand configurations as well as regularized model-fitting, assisting hand tracking and extracting handshape features. We construct SA images representing t...
متن کامل